Metric learning for kernel ridge regression: assessment of molecular similarity
نویسندگان
چکیده
Abstract Supervised and unsupervised kernel-based algorithms widely used in the physical sciences depend upon notion of similarity . Their reliance on pre-defined distance metrics—e.g. Euclidean or Manhattan distance—are problematic especially when combination with high-dimensional feature vectors for which measure does not well-reflect differences target property. Metric learning is an elegant approach to surmount this shortcoming find a property-informed transformation space. We propose new algorithm metric specifically adapted kernel ridge regression (KRR): (MLKRR). It based Learning Kernel Regression framework using Nadaraya-Watson estimator, we show be inferior KRR estimator typical physics-based machine tasks. The MLKRR allows superior predictive performance benchmark task atomisation energies QM9 molecules, as well generating more meaningful low-dimensional projections modified
منابع مشابه
Metric Learning for Kernel Regression
Kernel regression is a well-established method for nonlinear regression in which the target value for a test point is estimated using a weighted average of the surrounding training samples. The weights are typically obtained by applying a distance-based kernel function to each of the samples, which presumes the existence of a well-defined distance metric. In this paper, we construct a novel alg...
متن کاملSimilarity Metric Learning for aVariable - Kernel Classi
Nearest-neighbour interpolation algorithms have many useful properties for applications to learning, but they often exhibit poor generalization. In this paper, it is shown that much better generalization can be obtained by using a variable interpolation kernel in combination with conjugate gradient optimization of the similarity metric and kernel size. The resulting method is called variable-ke...
متن کاملKernel Regression with Sparse Metric Learning
Kernel regression is a popular non-parametric fitting technique. It aims at learning a function which estimates the targets for test inputs as precise as possible. Generally, the function value for a test input is estimated by a weighted average of the surrounding training examples. The weights are typically computed by a distancebased kernel function and they strongly depend on the distances b...
متن کاملSimilarity Metric Learning for a Variable-kernel Classiier
Nearest-neighbour interpolation algorithms have many useful properties for applications to learning, but they often exhibit poor generalization. In this paper, it is shown that much better generalization can be obtained by using a variable interpolation kernel in combination with conjugate gradient optimization of the similarity metric and kernel size. The resulting method is called variable-ke...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Machine learning: science and technology
سال: 2022
ISSN: ['2632-2153']
DOI: https://doi.org/10.1088/2632-2153/ac8e4f